Discrete MDL Predicts in Total Variation
نویسنده
چکیده
The Minimum Description Length (MDL) principle selects the model that has the shortest code for data plus model. We show that for a countable class of models, MDL predictions are close to the true distribution in a strong sense. The result is completely general. No independence, ergodicity, stationarity, identifiability, or other assumption on the model class need to be made. More formally, we show that for any countable class of models, the distributions selected by MDL (or MAP) asymptotically predict (merge with) the true measure in the class in total variation distance. Implications for non-i.i.d. domains like time-series forecasting, discriminative learning, and reinforcement learning are discussed.
منابع مشابه
The comparison of two high-order semi-discrete central schemes for solving hyperbolic conservation laws
This work presents two high-order, semi-discrete, central-upwind schemes for computing approximate solutions of 1D systems of conservation laws. We propose a central weighted essentially non-oscillatory (CWENO) reconstruction, also we apply a fourth-order reconstruction proposed by Peer et al., and afterwards, we combine these reconstructions with a semi-discrete central-upwind numerical flux ...
متن کاملA programmable 3.2-GOPS merged DRAM logic for video signal processing
This paper proposes a programmable high-performance architecture of datapath in the merged DRAM logic (MDL) for video signal processing. A model of a datapath in the programmable MDL is generated, and two basic parameters, total required clock cycles (TRCC) and DRAM access rate (DAR), are defined by analysis of the model. Design guidelines are suggested for the optimized video signal processor ...
متن کاملComputationally Efficient Methods for MDL-Optimal Density Estimation and Data Clustering
The Minimum Description Length (MDL) principle is a general, well-founded theoretical formalization of statistical modeling. The most important notion of MDL is the stochastic complexity, which can be interpreted as the shortest description length of a given sample of data relative to a model class. The exact definition of the stochastic complexity has gone through several evolutionary steps. T...
متن کاملStrong Asymptotic Assertions for Discrete MDL in Regression and Classification
We study the properties of the MDL (or maximum penalized complexity) estimator for Regression and Classification, where the underlying model class is countable. We show in particular a finite bound on the Hellinger losses under the only assumption that there is a “true” model contained in the class. This implies almost sure convergence of the predictive distribution to the true one at a fast ra...
متن کاملNML Computation Algorithms for Tree-Structured Multinomial Bayesian Networks
Typical problems in bioinformatics involve large discrete datasets. Therefore, in order to apply statistical methods in such domains, it is important to develop efficient algorithms suitable for discrete data. The minimum description length (MDL) principle is a theoretically well-founded, general framework for performing statistical inference. The mathematical formalization of MDL is based on t...
متن کامل